Neural Speed Reading via Skim-RNN

نویسندگان

  • Min Joon Seo
  • Sewon Min
  • Ali Farhadi
  • Hannaneh Hajishirzi
چکیده

Inspired by the principles of speed reading, we introduce Skim-RNN, a recurrent neural network (RNN) that dynamically decides to update only a small fraction of the hidden state for relatively unimportant input tokens. Skim-RNN gives computational advantage over an RNN that always updates the entire hidden state. Skim-RNN uses the same input and output interfaces as a standard RNN and can be easily used instead of RNNs in existing models. In our experiments, we show that Skim-RNN can achieve significantly reduced computational cost without losing accuracy compared to standard RNNs across five different natural language tasks. In addition, we demonstrate that the trade-off between accuracy and speed of Skim-RNN can be dynamically controlled during inference time in a stable manner. Our analysis also shows that Skim-RNN running on a single CPU offers lower latency compared to standard RNNs on GPUs.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Fast and Accurate Text Classification: Skim- Ming, Rereading and Early Stopping

Recent advances in recurrent neural nets (RNNs) have shown much promise in many applications in natural language processing. For most of these tasks, such as sentiment analysis of customer reviews, a recurrent neural net model parses the entire review before forming a decision. We argue that reading the entire input is not always necessary in practice, since a lot of reviews are often easy to c...

متن کامل

Coarse-to-Fine Question Answering for Long Documents

We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-ofthe-art models. While most successful approaches for reading comprehension rely on recurrent neural networks (RNNs), running them over long documents is prohibitively slow because it is difficult to parallelize over sequences. Inspired by how pe...

متن کامل

Learning Chaotic Dynamics using Tensor Recurrent Neural Networks

We present Tensor-RNN, a novel RNN architecture for multivariate forecasting in chaotic dynamical systems. Our proposed architecture captures highly nonlinear dynamic behavior by using high-order Markov states and transition functions. Furthermore, we decompose the highdimensional structure of the model using tensortrain networks to reduce the number of parameters while preserving the model per...

متن کامل

Fast reciprocal nearest neighbors clustering

This paper presents a novel approach for accelerating the popular Reciprocal Nearest Neighbors (RNN) clustering algorithm, i.e. the fast-RNN. We speed up the nearest neighbor chains construction via a novel dynamic slicing strategy for the projection search paradigm. We detail an efficient implementation of the clustering algorithm along with a novel data structure, and present extensive experi...

متن کامل

A New Hybrid-parameter Recurrent Neural Networks for Online Handwritten Chinese Character Recognition

The recurrent neural network (RNN) is appropriate for dealing with temporal sequences. In this paper, we present a deep RNN with new features and apply it for online handwritten Chinese character recognition. Compared with the existing RNN models, three innovations are involved in the proposed system. First, a new hidden layer function for RNN is proposed for learning temporal information bette...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1711.02085  شماره 

صفحات  -

تاریخ انتشار 2017